Toward Learning Gaussian Mixtures with Arbitrary Separation
نویسندگان
چکیده
In recent years analysis of complexity of learning Gaussian mixture models from sampled data has received significant attention in computational machine learning and theory communities. In this paper we present the first result showing that polynomial time learning of multidimensional Gaussian Mixture distributions is possible when the separation between the component means is arbitrarily small. Specifically, we present an algorithm for learning the parameters of a mixture of k identical spherical Gaussians in n-dimensional space with an arbitrarily small separation between the components, which is polynomial in dimension, inverse component separation and other input parameters for a fixed number of components k. The algorithm uses a projection to k dimensions and then a reduction to the 1-dimensional case. It relies on a theoretical analysis showing that two 1-dimensional mixtures whose densities are close in the L norm must have similar means and mixing coefficients. To produce the necessary lower bound for the L norm in terms of the distances between the corresponding means, we analyze the behavior of the Fourier transform of a mixture of Gaussians in one dimension around the origin, which turns out to be closely related to the properties of the Vandermonde matrix obtained from the component means. Analysis of minors of the Vandermonde matrix together with basic function approximation results allows us to provide a lower bound for the norm of the mixture in the Fourier domain and hence a bound in the original space. Additionally, we present a separate argument for reconstructing variance.
منابع مشابه
Learning Gaussian Mixtures with Arbitrary Separation
In recent years analysis of complexity of learning Gaussian mixture models from sampled datahas received significant attention in computational machine learning and theory communities. Inthis paper we present the first result showing that polynomial time learning of multidimensionalGaussian Mixture distributions is possible when the separation between the component means isarbit...
متن کاملA Neural Network for the Blind Separation of Non-Gaussian Sources
| In this paper, a two{layer neural network is presented that organizes itself to perform blind source separation. The inputs to the network are prewhitened linear mixtures of unknown independent source signals. An unsu-pervised nonlinear hebbian learning rule is used for training the network. After convergence, the network is able to extract the source signals from the mixtures, provided that ...
متن کاملLinear Time Clustering for High Dimensional Mixtures of Gaussian Clouds
Clustering mixtures of Gaussian distributions is a fundamental and challenging problem that is ubiquitous in various high-dimensional data processing tasks. While state-of-the-art work on learning Gaussian mixture models has focused primarily on improving separation bounds and their generalization to arbitrary classes of mixture models, less emphasis has been paid to practical computational eff...
متن کاملMinimax Theory for High-dimensional Gaussian Mixtures with Sparse Mean Separation
While several papers have investigated computationally and statistically efficient methods for learning Gaussian mixtures, precise minimax bounds for their statistical performance as well as fundamental limits in high-dimensional settings are not well-understood. In this paper, we provide precise information theoretic bounds on the clustering accuracy and sample complexity of learning a mixture...
متن کاملAn EM Algorithm for Independent Component Analysis in the Presence of Gaussian Noise
Abstract—An expectation-maximization (EM) algorithm for independent component analysis in the presence of gaussian noise is presented. The estimation of the conditional moments of the source posterior can be accomplished by maximum a posteriori estimation. The approximate conditional moments enable the development of an EM algorithm for inferring the most probable sources and learning the param...
متن کامل